Skip to content

[wip] deepseek r1 distilled llama 70B #1050

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from
Closed

Conversation

pawalt
Copy link
Member

@pawalt pawalt commented Jan 20, 2025

Keeping this as a WIP since we probably want to keep the Llama example. Only changes are:

  • vLLM version
  • Number of GPUs (full precision 70B requires 2xH100)
  • Chat format argument required in new vLLM version

@charlesfrye
Copy link
Collaborator

🚀 The docs preview is ready! Check it out here: https://modal-labs-examples--frontend-preview-dcedce0.modal.run

@luiscape
Copy link
Member

@pawalt looking forward to this one. I have a feeling that fine-tuning R1 would be great for us given its inference ROI.

@charlesfrye
Copy link
Collaborator

nice! I think we'll keep the smaller model in the main example, since the download time and cold boots for a 70B model are pretty painful.

@salman1993
Copy link

salman1993 commented Jan 26, 2025

@pawalt have you looked into enabling tool calling by any chance?
https://github.com/vllm-project/vllm/blob/6d0e3d372446cde48b387d4d3530e25fc6e06320/vllm/entrypoints/openai/serving_chat.py#L47

i get back responses from the API when i add in enable_auto_tools and tool_parser params, but tool calling doesn't work that well. my guess is the granite parser doesn't work well for deepseek models - we'd need the tool template it was trained with. also tried llama3_json parser but that doesn't work either

api_server.chat = lambda s: OpenAIServingChat(
        engine,
        ...
        chat_template_content_format="raw",
        enable_auto_tools=True,
        tool_parser="granite",
    )

example response i get (should have used get_weather tool call):

{
  "id": "chatcmpl-52b3d25b480144ff9dfecebf8e739d0c",
  "object": "chat.completion",
  "created": 1737908244,
  "model": "DeepSeek-R1-Distill-Llama-70B",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "<think>\n\n</think>\n\nHi there! I suggest getting online to get real-time information. If you have any other questions, please don't hesitate to let me know!",
        "tool_calls": []
      },
      "logprobs": null,
      "finish_reason": "stop",
      "stop_reason": null
    }
  ],
  ...
}

@charlesfrye
Copy link
Collaborator

closing since we now have a DeepSeek-R1 example (cf #1057)

@github-actions github-actions bot deleted the pawalt/deepseek_r1 branch March 1, 2025 15:35
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants